Improving Gradient Flow with Unrolled Highway Expectation Maximization

نویسندگان

چکیده

Integrating model-based machine learning methods into deep neural architectures allows one to leverage both the expressive power of nets and ability incorporate domain-specific knowledge. In particular, many works have employed expectation maximization (EM) algorithm in form an unrolled layer-wise structure that is jointly trained with a backbone network. However, it difficult discriminatively train network by backpropagating through EM iterations as they are prone vanishing gradient problem. To address this issue, we propose Highway Expectation Maximization Networks (HEMNet), which comprised generalized (GEM) based on Newton-Rahpson method. HEMNet features scaled skip connections, or highways, along depths architecture, resulting improved flow during backpropagation while incurring negligible additional computation memory costs compared standard EM. Furthermore, preserves underlying procedure, thereby fully retaining convergence properties original algorithm. We achieve significant improvement performance several semantic segmentation benchmarks empirically show effectively alleviates decay.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

High-Dimensional Variance-Reduced Stochastic Gradient Expectation-Maximization Algorithm

We propose a generic stochastic expectationmaximization (EM) algorithm for the estimation of high-dimensional latent variable models. At the core of our algorithm is a novel semi-stochastic variance-reduced gradient designed for the Qfunction in the EM algorithm. Under a mild condition on the initialization, our algorithm is guaranteed to attain a linear convergence rate to the unknown paramete...

متن کامل

Expectation Maximization

The Expectation Maximization (EM) algorithm [1, 2] is one of the most widely used algorithms in statistics. Suppose we are given some observed data X and a model family parametrized by θ, and would like to find the θ which maximizes p(X |θ), i.e. the maximum likelihood estimator. The basic idea of EM is actually quite simple: when direct maximization of p(X |θ) is complicated we can augment the...

متن کامل

Autonomous Exploration with Expectation-Maximization

We consider the problem of autonomous mobile robot exploration in an unknown environment for the purpose of building an accurate feature-based map efficiently. Most literature on this subject is focused on the combination of a variety of utility functions, such as curbing robot pose uncertainty and the entropy of occupancy grid maps. However, the effect of uncertain poses is typically not well ...

متن کامل

Data Synthesis with Expectation-Maximization

A problem of increasing importance in computer graphics is to generate data with the style of some previous training data, but satisfying new constraints. If we use a probabilistic latent variable model, then learning the model will normally be performed using Expectation-Maximization (EM), or one of its generalizations. We show that data synthesis for such problems can also be performed using ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2021

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v35i11.17167